Goto

Collaborating Authors

 Summit County


A Unifying Human-Centered AI Fairness Framework

Rahman, Munshi Mahbubur, Pan, Shimei, Foulds, James R.

arXiv.org Artificial Intelligence

The increasing use of Artificial Intelligence (AI) in critical societal domains has amplified concerns about fairness, particularly regarding unequal treatment across sensitive attributes such as race, gender, and socioeconomic status. While there has been substantial work on ensuring AI fairness, navigating trade-offs between competing notions of fairness as well as predictive accuracy remains challenging, creating barriers to the practical deployment of fair AI systems. To address this, we introduce a unifying human-centered fairness framework that systematically covers eight distinct fairness metrics, formed by combining individual and group fairness, infra-marginal and intersectional assumptions, and outcome-based and equality-of-opportunity (EOO) perspectives. This structure allows stakeholders to align fairness interventions with their values and contextual considerations. The framework uses a consistent and easy-to-understand formulation for all metrics to reduce the learning curve for non-experts. Rather than privileging a single fairness notion, the framework enables stakeholders to assign weights across multiple fairness objectives, reflecting their priorities and facilitating multi-stakeholder compromises. We apply this approach to four real-world datasets: the UCI Adult census dataset for income prediction, the COMPAS dataset for criminal recidivism, the German Credit dataset for credit risk assessment, and the MEPS dataset for healthcare utilization. We show that adjusting weights reveals nuanced trade-offs between different fairness metrics. Finally, through case studies in judicial decision-making and healthcare, we demonstrate how the framework can inform practical and value-sensitive deployment of fair AI systems.


General Catalyst CEO Hemant Taneja on Aligning Profit With Purpose

TIME - Tech

Booth is a reporter at TIME. Hemant Taneja, CEO, General Catalyst speaks on stage during The Summit on U.S. Resilience hosted by General Catalyst Institute at The Salamander on Nov. 17, 2025 in Washington, DC. Hemant Taneja, CEO, General Catalyst speaks on stage during The Summit on U.S. Resilience hosted by General Catalyst Institute at The Salamander on Nov. 17, 2025 in Washington, DC. Booth is a reporter at TIME. Hemant Taneja, who leads one of the world's largest venture firms, believes doing good isn't just the right thing to do.


Shocking video you MUST watch before voting for Mamdani: Here's what will become of NYC under him... and it's worse than everyone fears

Daily Mail - Science & tech

Stunning before-and-after photos show the seven most dramatic changes in Trump's controversial White House makeover She was a respected Teacher of the Year finalist... until she lost everything when Charlie Kirk was killed. Inside Andrew's family summit: How Fergie wailed and'melted down' at title loss, Beatrice and Eugenie were'blindsided' and now daughters' assets face'ethics check' to avoid more scandal: BARBARA DAVIES I have no sympathy for Britney Spears. What if her latest stunt had killed a kid? It's time to admit the truth about this public menace: KENNEDY'Nazi texts' leakers UNMASKED: Alleged White House saboteurs are finally exposed... and so is their twisted motive for destroying political prodigy Extraordinary story behind GM's decision to ax much-loved CarPlay... and sinister reason ALL manufacturers will follow What is Charcot-Marie-Tooth disease... the devastating condition that killed 9-1-1 Nashville actor Isabelle Tate Bijou Phillips files to change daughter's name after ex Danny Masterson's rape conviction Treasure hunters seeking Nazi gold worth £200MILLION believe they have'found the real thing' after'monumental' discovery under remains of SS palace'brothel' Former Gambino mob boss'Sammy the Bull' Gravano reveals the truth behind the NBA betting scandal My wife won't get a job and I feel broken trying to provide for our family. Hold on, says DEAR CAROLINE... that's bad enough but your letter raises a MUCH bigger red flag I got the body of my dreams at 51 by following 9 simple rules, says beauty guru ROSIE GREEN.


Leveraging Vulnerabilities in Temporal Graph Neural Networks via Strategic High-Impact Assaults

Jeon, Dong Hyun, Zhu, Lijing, Li, Haifang, Li, Pengze, Feng, Jingna, Duan, Tiehang, Song, Houbing Herbert, Tao, Cui, Niu, Shuteng

arXiv.org Artificial Intelligence

Temporal Graph Neural Networks (TGNNs) have become indispensable for analyzing dynamic graphs in critical applications such as social networks, communication systems, and financial networks. However, the robustness of TGNNs against adversarial attacks, particularly sophisticated attacks that exploit the temporal dimension, remains a significant challenge. Existing attack methods for Spatio-Temporal Dynamic Graphs (STDGs) often rely on simplistic, easily detectable perturbations (e.g., random edge additions/deletions) and fail to strategically target the most influential nodes and edges for maximum impact. We introduce the High Impact Attack (HIA), a novel restricted black-box attack framework specifically designed to overcome these limitations and expose critical vulnerabilities in TGNNs. HIA leverages a data-driven surrogate model to identify structurally important nodes (central to network connectivity) and dynamically important nodes (critical for the graph's temporal evolution). It then employs a hybrid perturbation strategy, combining strategic edge injection (to create misleading connections) and targeted edge deletion (to disrupt essential pathways), maximizing TGNN performance degradation. Importantly, HIA minimizes the number of perturbations to enhance stealth, making it more challenging to detect. Comprehensive experiments on five real-world datasets and four representative TGNN architectures (TGN, JODIE, DySAT, and TGAT) demonstrate that HIA significantly reduces TGNN accuracy on the link prediction task, achieving up to a 35.55% decrease in Mean Reciprocal Rank (MRR) - a substantial improvement over state-of-the-art baselines. These results highlight fundamental vulnerabilities in current STDG models and underscore the urgent need for robust defenses that account for both structural and temporal dynamics.


Ensembling Multilingual Transformers for Robust Sentiment Analysis of Tweets

Bilehsavar, Meysam Shirdel, Mahmoudi, Negin, Torkamani, Mohammad Jalili, Kiashemshaki, Kiana

arXiv.org Artificial Intelligence

Sentiment analysis is a very important natural language processing activity in which one identifies the polarity of a text, whether it conveys positive, negative, or neutral sentiment. Along with the growth of social media and the Internet, the significance of sentiment analysis has grown across numerous industries such as marketing, politics, and customer service. Sentiment analysis is flawed, however, when applied to foreign languages, particularly when there is no labelled data to train models upon. In this study, we present a transformer ensemble model and a large language model (LLM) that employs sentiment analysis of other languages. We used multi languages dataset. Sentiment was then assessed for sentences using an ensemble of pre-trained sentiment analysis models: bert-base-multilingual-uncased-sentiment, and XLM-R. Our experimental results indicated that sentiment analysis performance was more than 86% using the proposed method.


LLM-Driven Adaptive 6G-Ready Wireless Body Area Networks: Survey and Framework

Torkamani, Mohammad Jalili, Mahmoudi, Negin, Kiashemshaki, Kiana

arXiv.org Artificial Intelligence

--Wireless Body Area Networks (WBANs) enable continuous monitoring of physiological signals for applications ranging from chronic disease management to emergency response. Recent advances in 6G communications, post-quantum cryptography, and energy harvesting have the potential to enhance WBAN performance. However, integrating these technologies into a unified, adaptive system remains a challenge. We propose a novel Large Language Model-driven adaptive WBAN framework in which a Large Language Model acts as a cognitive control plane, coordinating routing, physical layer selection, micro-energy harvesting, and post-quantum security in real time. Our review highlights the limitations of current heuristic-based designs and outlines a research agenda for resource-constrained, 6G-ready medical systems. This approach aims to enable ultra-reliable, secure, and self-optimizing WBANs for next-generation mobile health applications.


LinkAnchor: An Autonomous LLM-Based Agent for Issue-to-Commit Link Recovery

Akhavan, Arshia, Hosseinpour, Alireza, Heydarnoori, Abbas, Keshani, Mehdi

arXiv.org Artificial Intelligence

--Issue-to-commit link recovery plays an important role in software traceability and improves project management. However, it remains a challenging task. A study on GitHub shows that only 42.2% of the issues are correctly linked to their commits. This highlights the potential for further development and research in this area. Existing studies have employed various AI/ML-based approaches, and with the recent development of large language models, researchers have leveraged LLMs to tackle this problem. These approaches suffer from two main issues. First, LLMs are constrained by limited context windows and cannot ingest all of the available data sources, such as long commit histories, extensive issue comments, and large code repositories. Second, most methods operate on individual issue-commit pairs; That is, given a single issue-commit pair, they determine whether the commit resolves the issue. This quickly becomes impractical in real-world repositories containing tens of thousands of commits. T o address these limitations, we present LinkAnchor, the first autonomous LLM-based agent designed for issue-to-commit link recovery. The lazy-access architecture of LinkAnchor enables the underlying LLM to access the rich context of software, spanning commits, issue comments, and code files, without exceeding the token limit by dynamically retrieving only the most relevant contextual data. Additionally, LinkAnchor is able to automatically pinpoint the target commit rather than exhaustively scoring every possible candidate. Our evaluations show that LinkAnchor outperforms state-of-the-art issue-to-commit link recovery approaches by 60-262% in Hit@1 score across all our case study projects. We also publicly release LinkAnchor [1] as a ready-to-use tool, along with our replication package. LinkAnchor is designed and tested for GitHub and Jira, and is easily extendable to other platforms. Trace link recovery (TLR) is the process of identifying and establishing connections between related software artifacts, such as requirements, code, tests, and documentation.


Hallucinations and Key Information Extraction in Medical Texts: A Comprehensive Assessment of Open-Source Large Language Models

Das, Anindya Bijoy, Ahmed, Shibbir, Sakib, Shahnewaz Karim

arXiv.org Artificial Intelligence

Clinical summarization is crucial in healthcare as it distills complex medical data into digestible information, enhancing patient understanding and care management. Large language models (LLMs) have shown significant potential in automating and improving the accuracy of such summarizations due to their advanced natural language understanding capabilities. These models are particularly applicable in the context of summarizing medical/clinical texts, where precise and concise information transfer is essential. In this paper, we investigate the effectiveness of open-source LLMs in extracting key events from discharge reports, including admission reasons, major in-hospital events, and critical follow-up actions. In addition, we also assess the prevalence of various types of hallucinations in the summaries produced by these models. Detecting hallucinations is vital as it directly influences the reliability of the information, potentially affecting patient care and treatment outcomes. We conduct comprehensive simulations to rigorously evaluate the performance of these models, further probing the accuracy and fidelity of the extracted content in clinical summarization. Our results reveal that while the LLMs (e.g., Qwen2.5 and DeepSeek-v2) perform quite well in capturing admission reasons and hospitalization events, they are generally less consistent when it comes to identifying follow-up recommendations, highlighting broader challenges in leveraging LLMs for comprehensive summarization.


Trustworthy Medical Imaging with Large Language Models: A Study of Hallucinations Across Modalities

Das, Anindya Bijoy, Sakib, Shahnewaz Karim, Ahmed, Shibbir

arXiv.org Artificial Intelligence

Large Language Models (LLMs) are increasingly applied to medical imaging tasks, including image interpretation and synthetic image generation. However, these models often produce hallucinations, which are confident but incorrect outputs that can mislead clinical decisions. This study examines hallucinations in two directions: image to text, where LLMs generate reports from X-ray, CT, or MRI scans, and text to image, where models create medical images from clinical prompts. We analyze errors such as factual inconsistencies and anatomical inaccuracies, evaluating outputs using expert informed criteria across imaging modalities. Our findings reveal common patterns of hallucination in both interpretive and generative tasks, with implications for clinical reliability. We also discuss factors contributing to these failures, including model architecture and training data. By systematically studying both image understanding and generation, this work provides insights into improving the safety and trustworthiness of LLM driven medical imaging systems.


Enhancing Glass Defect Detection with Diffusion Models: Addressing Imbalanced Datasets in Manufacturing Quality Control

Boroujeni, Sajjad Rezvani, Abedi, Hossein, Bush, Tom

arXiv.org Artificial Intelligence

Visual defect detection in industrial glass manufacturing remains a critical challenge due to the low frequency of defective products, leading to imbalanced datasets that limit the performance of deep learning models and computer vision systems. This paper presents a novel approach using Denoising Diffusion Probabilistic Models (DDPMs) to generate synthetic defective glass product images for data augmentation, effectively addressing class imbalance issues in manufacturing quality control and automated visual inspection. The methodology significantly enhances image classification performance of standard CNN architectures (ResNet50V2, EfficientNetB0, and MobileNetV2) in detecting anomalies by increasing the minority class representation. Experimental results demonstrate substantial improvements in key machine learning metrics, particularly in recall for defective samples across all tested deep neural network architectures while maintaining perfect precision on the validation set. The most dramatic improvement was observed in ResNet50V2's overall classification accuracy, which increased from 78\% to 93\% when trained with the augmented data. This work provides a scalable, cost-effective approach to enhancing automated defect detection in glass manufacturing that can potentially be extended to other industrial quality assurance systems and industries with similar class imbalance challenges.